منابع مشابه
A continuous Newton-type method for unconstrained optimization
In this paper, we propose a continuous Newton-type method in the form of an ordinary differential equation by combining the negative gradient and Newton’s direction. It is shown that for a general function f(x), our method converges globally to a connected subset of the stationary points of f(x) under some mild conditions; and converges globally to a single stationary point for a real analytic ...
متن کاملCommunication-Efficient Distributed Optimization using an Approximate Newton-type Method
We present a novel Newton-type method for distributed optimization, which is particularly well suited for stochastic optimization and learning problems. For quadratic objectives, the method enjoys a linear rate of convergence which provably improves with the data size, requiring an essentially constant number of iterations under reasonable assumptions. We provide theoretical and empirical evide...
متن کاملAn Approximate Quasi-Newton Bundle-Type Method for Nonsmooth Optimization
and Applied Analysis 3 rate of convergence under some additional assumptions, and it should be noted that we only use the approximate values of the objective function and its subgradients which makes the algorithm easier to implement. Some notations are listed below for presenting the algorithm. (i) ∂f(x) = {ξ ∈ Rn | f(z) ≥ f(x) + ξT(z − x), ∀z ∈ R}, the subdifferential of f at x, and each such...
متن کاملRobust Parallel Newton { Multilevel
The present paper is devoted to the numerical solution of nonlinear boundary value problems arising in the magnetic eld computation and in solid mechanics. These problems are discretized by using nite elements. The nonlinearity is handled by a nested Newton solver, and the linear systems of algebraic equations within each Newton step are solved by means of various iterative solvers, namely mult...
متن کاملRegularized Newton method for unconstrained convex optimization
We introduce the regularized Newton method (rnm) for unconstrained convex optimization. For any convex function, with a bounded optimal set, the rnm generates a sequence that converges to the optimal set from any starting point. Moreover the rnm requires neither strong convexity nor smoothness properties in the entire space. If the function is strongly convex and smooth enough in the neighborho...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Optimization Methods and Software
سال: 2019
ISSN: 1055-6788,1029-4937
DOI: 10.1080/10556788.2019.1700256